Tootfinder

Opt-in global Mastodon full text search. Join the index!

No exact results. Similar results found.
@arXiv_csCV_bot@mastoxiv.page
2024-05-10 08:29:32

This arxiv.org/abs/2402.18573 has been replaced.
initial toot: mastoxiv.page/@arXiv_csCV_…

@arXiv_csLG_bot@mastoxiv.page
2024-04-10 06:51:06

Unifying Low Dimensional Observations in Deep Learning Through the Deep Linear Unconstrained Feature Model
Connall Garrod, Jonathan P. Keating
arxiv.org/abs/2404.06106

@arXiv_physicsfludyn_bot@mastoxiv.page
2024-04-09 09:03:18

This arxiv.org/abs/2402.04690 has been replaced.
link: scholar.google.com/scholar?q=a

@cowboys@darktundra.xyz
2024-05-04 12:44:37

Flashback 1975: The Packers Shocked the Cowboys as Bart Starr Earned His First Win as Head Coach yardbarker.com/nfl/articles/fl

@arXiv_astrophIM_bot@mastoxiv.page
2024-05-08 07:00:00

Optical Photon Emission in Extended Airshowers -- Hybrid computing in the context of CORSIKA 8
Dominik Baack
arxiv.org/abs/2405.04229

@arXiv_mathDG_bot@mastoxiv.page
2024-04-09 08:51:18

This arxiv.org/abs/2211.07762 has been replaced.
link: scholar.google.com/scholar?q=a

@cheryanne@aus.social
2024-04-30 22:00:59

Unbuttoned Podcast
Great Australian Pods Podcast Directory: #GreatAusPods

Unbuttoned Podcast
Screenshot of the podcast listing on the Great Australian Pods website
@arXiv_csHC_bot@mastoxiv.page
2024-05-01 07:17:25

Fake it to make it: Using synthetic data to remedy the data shortage in joint multimodal speech-and-gesture synthesis
Shivam Mehta, Anna Deichler, Jim O'Regan, Birger Mo\"ell, Jonas Beskow, Gustav Eje Henter, Simon Alexanderson
arxiv.org/abs/2404.19622 arxiv.org/pdf/2404.19622
arXiv:2404.19622v1 Announce Type: new
Abstract: Although humans engaged in face-to-face conversation simultaneously communicate both verbally and non-verbally, methods for joint and unified synthesis of speech audio and co-speech 3D gesture motion from text are a new and emerging field. These technologies hold great promise for more human-like, efficient, expressive, and robust synthetic communication, but are currently held back by the lack of suitably large datasets, as existing methods are trained on parallel data from all constituent modalities. Inspired by student-teacher methods, we propose a straightforward solution to the data shortage, by simply synthesising additional training material. Specifically, we use unimodal synthesis models trained on large datasets to create multimodal (but synthetic) parallel training data, and then pre-train a joint synthesis model on that material. In addition, we propose a new synthesis architecture that adds better and more controllable prosody modelling to the state-of-the-art method in the field. Our results confirm that pre-training on large amounts of synthetic data improves the quality of both the speech and the motion synthesised by the multimodal model, with the proposed architecture yielding further benefits when pre-trained on the synthetic data. See shivammehta25.github.io/MAGI/ for example output.

@arXiv_csSE_bot@mastoxiv.page
2024-04-29 06:52:59

A Unified Debugging Approach via LLM-Based Multi-Agent Synergy
Cheryl Lee, Chunqiu Steven Xia, Jen-tse Huang, Zhouruixin Zhu, Lingming Zhang, Michael R. Lyu
arxiv.org/abs/2404.17153

@arXiv_csHC_bot@mastoxiv.page
2024-04-04 07:20:27

A Unified Editing Method for Co-Speech Gesture Generation via Diffusion Inversion
Zeyu Zhao, Nan Gao, Zhi Zeng, Guixuan Zhang, Jie Liu, Shuwu Zhang
arxiv.org/abs/2404.02411